Qualitative Possibilistic Mixed-Observable MDPs
نویسندگان
چکیده
Possibilistic and qualitative POMDPs (πPOMDPs) are counterparts of POMDPs used to model situations where the agent’s initial belief or observation probabilities are imprecise due to lack of past experiences or insufficient data collection. However, like probabilistic POMDPs, optimally solving πPOMDPs is intractable: the finite belief state space exponentially grows with the number of system’s states. In this paper, a possibilistic version of Mixed-Observable MDPs is presented to get around this issue: the complexity of solving π-POMDPs, some state variables of which are fully observable, can be then dramatically reduced. A value iteration algorithm for this new formulation under infinite horizon is next proposed and the optimality of the returned policy (for a specified criterion) is shown assuming the existence of a ”stay” action in some goal states. Experimental work finally shows that this possibilistic model outperforms probabilistic POMDPs commonly used in robotics, for a target recognition problem where the agent’s observations are imprecise.
منابع مشابه
Structured Possibilistic Planning Using Decision Diagrams
Qualitative Possibilistic Mixed-Observable MDPs (πMOMDPs), generalizing π-MDPs and π-POMDPs, are well-suited models to planning under uncertainty with mixed-observability when transition, observation and reward functions are not precisely known and can be qualitatively described. Functions defining the model as well as intermediate calculations are valued in a finite possibilistic scale L, whic...
متن کاملEfficient Policies for Stationary Possibilistic Markov Decision Processes
Possibilistic Markov Decision Processes offer a compact and tractable way to represent and solve problems of sequential decision under qualitative uncertainty. Even though appealing for its ability to handle qualitative problems, this model suffers from the drowning effect that is inherent to possibilistic decision theory. The present paper proposes to escape the drowning effect by extending to...
متن کاملEmpirical Comparison of Probabilistic and Possibilistic Markov Decision Processes Algorithms
Classical stochastic Markov Decision Processes (MDPs) and possibilistic MDPs ( -MDPs) aim at solving the same kind of problems, involving sequential decision making under uncertainty. The underlying uncertainty model (probabilistic / possibilistic) and preference model (reward / satisfaction degree) change, but the algorithms, based on dynamic programming, are similar. So, a question maybe rais...
متن کاملA Possibilistic Model for Qualitative Sequential Decision Problems under Uncertainty in Partially Observable Environments
In this article we propose a qualitative ( ordi nal) counterpart for the Partially Observable Markov Decision Processes rnodel (POMDP) in which the uncertainty, as well as the prefer ences of the agent, are modeled by possibility distributions. This qualitative counterpart of the POMDP model relies on a possibilistic theory of decision under uncertainty, recently developed. One advantage of s...
متن کاملQualitative MDPs and POMDPs: An Order-Of-Magnitude Approximation
We develop a qualitative theory of Markov Decision Processes (mdps) and Partially Observable mdps that can be used to model sequential decision making tasks when only qualitative information is available. Our approach is based upon an order-of-magnitude approximation of both probabilities and utilities, similar to 2-semantics. The result is a qualitative theory that has close ties with the stan...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1309.6826 شماره
صفحات -
تاریخ انتشار 2013